A Modular Approach to Knowledge Transfer Between Reinforcement Learning Agents
نویسنده
چکیده
Reinforcement learning is a general approach to learning reactive control policies. It is an unsupervised learning technique, making it a candidate for use in system that adapts to changing tasks and environment by autonomously devising a new strategy. Unfortunately, reinforcement learning methods are slow to converge to a solution, rendering them impractical in most cases. The key shortcoming of most reinforcement learning systems is that they lack the ability to incorporate prior knowledge into the learning process. This is wasteful of any previously learned information that is related to the target task. This paper investigates a scheme for incorporating such prior knowledge into a reinforcement learning agent. The methodology, called Skill Advice Guided Exploration (SAGE), allows multiple, possibly conflicting, sources of knowledge to be incorporated simultaneously. Knowledge is abstracted as action selection advice within this framework. Accordingly, no assumptions must be made about the internal representations of the information, allowing knowledge sources to be as diverse as a human and a robot. Furthermore, a SAGE-based system can adapt according to new knowledge acquired during learning and is robust to incorrect information. The potential of this methodology is demonstrated on a set of discrete learning tasks. Results are presented showing that a SAGE-based system given correct information can outperform other approaches and that incorrect knowledge does not prevent the task from being learned. The benefits and limitations of this work are discussed as are possible extensions * Prepared in partial fulfillment of the requirements for the degree of Master of Science in Electrical and Computer Engineering.
منابع مشابه
Hierarchical Functional Concepts for Knowledge Transfer among Reinforcement Learning Agents
This article introduces the notions of functional space and concept as a way of knowledge representation and abstraction for Reinforcement Learning agents. These definitions are used as a tool of knowledge transfer among agents. The agents are assumed to be heterogeneous; they have different state spaces but share a same dynamic, reward and action space. In other words, the agents are assumed t...
متن کاملLearning in Multi-agent Systems with Sparse Interactions by Knowledge Transfer and Game Abstraction
In many multi-agent systems, the interactions between agents are sparse and exploiting interaction sparseness in multiagent reinforcement learning (MARL) can improve the learning performance. Also, agents may have already learnt some single-agent knowledge (e.g., local value function) before the multi-agent learning process. In this work, we investigate how such knowledge can be utilized to lea...
متن کاملAn Unsupervised Learning Method for an Attacker Agent in Robot Soccer Competitions Based on the Kohonen Neural Network
RoboCup competition as a great test-bed, has turned to a worldwide popular domains in recent years. The main object of such competitions is to deal with complex behavior of systems whichconsist of multiple autonomous agents. The rich experience of human soccer player can be used as a valuable reference for a robot soccer player. However, because of the differences between real and simulated soc...
متن کاملTransfer Learning in Multi-Agent Reinforcement Learning Domains
Transfer learning refers to the process of reusing knowledge from past tasks in order to speed up the learning procedure in new tasks. In reinforcement learning, where agents often require a considerable amount of training, transfer learning comprises a suitable solution for speeding up learning. Transfer learning methods have primarily been applied in single-agent reinforcement learning algori...
متن کاملContext Transfer in Reinforcement Learning Using Action-Value Functions
This paper discusses the notion of context transfer in reinforcement learning tasks. Context transfer, as defined in this paper, implies knowledge transfer between source and target tasks that share the same environment dynamics and reward function but have different states or action spaces. In other words, the agents learn the same task while using different sensors and actuators. This require...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2000